我们的目标是将denoisis扩散隐式模型(DDIM)扩展到一般扩散模型〜(DMS)。我们没有像原始DDIM论文那样构建非马尔科夫no噪声过程,而是从数值的角度研究了DDIM的机制。我们发现,在求解相应的随机微分方程时,可以通过使用分数的一些特定近似值来获得DDIM。我们提出了DDIM加速效应的解释,该解释还解释了确定性抽样方案的优势,而不是随机采样方案进行快速采样。在此洞察力的基础上,我们将DDIM扩展到一般的DMS,并在参数化分数网络时进行了小而微妙的修改。当应用于批判性抑制的Langevin扩散模型时,最近提出的一种新型的扩散模型通过以速度增强扩散过程,我们的算法在CIFAR10上达到了2.28的FID分数,仅具有50个数量的得分功能评估(NFES)(NFES〜(NFES) )和仅有27个NFE的FID分数为2.87,比所有具有相同NFE的现有方法要好。代码可从https://github.com/qsh-zh/gddim获得
translated by 谷歌翻译
过去的几年见证了扩散模型〜(DMS)在生成建模任务中生成高保真样本方面取得的巨大成功。 DM的主要局限性是其臭名昭著的缓慢采样程序,通常需要数百到数千至数千个的时间离散步骤,以达到所需的准确性。我们的目标是为DMS开发快速采样方法,该方法的步骤少得多,同时保留了高样本质量。为此,我们系统地分析了DMS中的采样程序,并确定影响样本质量的关键因素,其中离散化方法至关重要。通过仔细检查学习的扩散过程,我们提出了扩散指数积分取样器〜(DEIS)。它基于设计用于离散的普通微分方程(ODE)的指数积分器,并利用学习扩散过程的半线性结构来减少离散误差。所提出的方法可以应用于任何DMS,并可以在短短10个步骤中生成高保真样本。在我们的实验中,一个A6000 GPU大约需要3分钟才能从CIFAR10产生$ 50K $的图像。此外,通过直接使用预训练的DMS,当得分函数评估的数量〜(NFE)的数量有限时,我们实现了最先进的采样性能,例如,使用10 NFES,3.37 FID和9.74的4.17 FID,仅为9.74 CIFAR10上的15个NFE。代码可从https://github.com/qsh-zh/deis获得
translated by 谷歌翻译
我们呈现路径积分采样器〜(PIS),一种新型算法,用于从非正规化概率密度函数中绘制样本。 PIS建立在SCHR \“odinger桥问题上,旨在恢复鉴于其初始分布和终端分布的扩散过程的最可能演变。PIS从初始分布中抽取样品,然后通过SCHR \”传播样本“少剂桥到达终端分布。应用Girsanov定理,通过简单的先前扩散,我们将PIS制定为随机最佳控制问题,其运行成本是根据目标分布选择控制能量和终端成本。通过将控件建模为神经网络,我们建立了一种可以训练结束到底的采样算法。在使用子最优控制时,我们在Wassersein距离方面提供了PIS的采样质量的理论典范。此外,路径积分理论用于计算样本的重要性权重,以补偿由控制器的次级最优性和时间离散化引起的偏差。我们通过关于各种任务的其他启动采样方法进行了实验证明了PIS的优势。
translated by 谷歌翻译
在本文中,我们提出了一种算法,用于估计聚合观测的时间均匀隐马尔可夫模型的参数。当只有每次步骤的个人数量的人口级别计数时,都会出现此问题,从中寻求学习单个隐藏的马尔可夫模型。我们的算法是在期望 - 最大化和最近提出的聚合推理算法,池中信念传播的建立。与现有方法相比,诸如具有非线性信念传播的期望最大化,我们的算法表现出收敛保证。此外,当记录与单个单独的观察时,我们的学习框架自然地降低了标准的BAUM-Welch学习算法。我们进一步扩展了我们的学习算法以处理具有连续观察的HMM。我们的算法的功效在各种数据集上进行了演示。
translated by 谷歌翻译
Masked image modeling (MIM) performs strongly in pre-training large vision Transformers (ViTs). However, small models that are critical for real-world applications cannot or only marginally benefit from this pre-training approach. In this paper, we explore distillation techniques to transfer the success of large MIM-based pre-trained models to smaller ones. We systematically study different options in the distillation framework, including distilling targets, losses, input, network regularization, sequential distillation, etc, revealing that: 1) Distilling token relations is more effective than CLS token- and feature-based distillation; 2) An intermediate layer of the teacher network as target perform better than that using the last layer when the depth of the student mismatches that of the teacher; 3) Weak regularization is preferred; etc. With these findings, we achieve significant fine-tuning accuracy improvements over the scratch MIM pre-training on ImageNet-1K classification, using all the ViT-Tiny, ViT-Small, and ViT-base models, with +4.2%/+2.4%/+1.4% gains, respectively. Our TinyMIM model of base size achieves 52.2 mIoU in AE20K semantic segmentation, which is +4.1 higher than the MAE baseline. Our TinyMIM model of tiny size achieves 79.6% top-1 accuracy on ImageNet-1K image classification, which sets a new record for small vision models of the same size and computation budget. This strong performance suggests an alternative way for developing small vision Transformer models, that is, by exploring better training methods rather than introducing inductive biases into architectures as in most previous works. Code is available at https://github.com/OliverRensu/TinyMIM.
translated by 谷歌翻译
In this paper, we propose a robust 3D detector, named Cross Modal Transformer (CMT), for end-to-end 3D multi-modal detection. Without explicit view transformation, CMT takes the image and point clouds tokens as inputs and directly outputs accurate 3D bounding boxes. The spatial alignment of multi-modal tokens is performed implicitly, by encoding the 3D points into multi-modal features. The core design of CMT is quite simple while its performance is impressive. CMT obtains 73.0% NDS on nuScenes benchmark. Moreover, CMT has a strong robustness even if the LiDAR is missing. Code will be released at https://github.com/junjie18/CMT.
translated by 谷歌翻译
Dataset distillation has emerged as a prominent technique to improve data efficiency when training machine learning models. It encapsulates the knowledge from a large dataset into a smaller synthetic dataset. A model trained on this smaller distilled dataset can attain comparable performance to a model trained on the original training dataset. However, the existing dataset distillation techniques mainly aim at achieving the best trade-off between resource usage efficiency and model utility. The security risks stemming from them have not been explored. This study performs the first backdoor attack against the models trained on the data distilled by dataset distillation models in the image domain. Concretely, we inject triggers into the synthetic data during the distillation procedure rather than during the model training stage, where all previous attacks are performed. We propose two types of backdoor attacks, namely NAIVEATTACK and DOORPING. NAIVEATTACK simply adds triggers to the raw data at the initial distillation phase, while DOORPING iteratively updates the triggers during the entire distillation procedure. We conduct extensive evaluations on multiple datasets, architectures, and dataset distillation techniques. Empirical evaluation shows that NAIVEATTACK achieves decent attack success rate (ASR) scores in some cases, while DOORPING reaches higher ASR scores (close to 1.0) in all cases. Furthermore, we conduct a comprehensive ablation study to analyze the factors that may affect the attack performance. Finally, we evaluate multiple defense mechanisms against our backdoor attacks and show that our attacks can practically circumvent these defense mechanisms.
translated by 谷歌翻译
Blind image quality assessment (BIQA) remains challenging due to the diversity of distortion and image content variation, which complicate the distortion patterns crossing different scales and aggravate the difficulty of the regression problem for BIQA. However, existing BIQA methods often fail to consider multi-scale distortion patterns and image content, and little research has been done on learning strategies to make the regression model produce better performance. In this paper, we propose a simple yet effective Progressive Multi-Task Image Quality Assessment (PMT-IQA) model, which contains a multi-scale feature extraction module (MS) and a progressive multi-task learning module (PMT), to help the model learn complex distortion patterns and better optimize the regression issue to align with the law of human learning process from easy to hard. To verify the effectiveness of the proposed PMT-IQA model, we conduct experiments on four widely used public datasets, and the experimental results indicate that the performance of PMT-IQA is superior to the comparison approaches, and both MS and PMT modules improve the model's performance.
translated by 谷歌翻译
Automatic music generation with artificial intelligence typically requires a large amount of data which is hard to obtain for many less common genres and musical instruments. To tackle this issue, we present ongoing work and preliminary findings on the possibility for deep models to transfer knowledge from language to music, by finetuning large language models pre-trained on a massive text corpus on only hundreds of MIDI files of drum performances. We show that by doing so, one of the largest, state-of-the-art models (GPT3) is capable of generating reasonable drum grooves, while models that are not pre-trained (Transformer) shows no such ability beyond naive repetition. Evaluating generated music is a challenging task, more so is evaluating drum grooves with little precedence in literature. Hence, we propose a tailored structural evaluation method and analyze drum grooves produced by GPT3 compared to those played by human professionals, exposing the strengths and weaknesses of such generation by language-to-music transfer. Our findings suggest that language-to-music transfer learning with large language models is viable and promising.
translated by 谷歌翻译
Few Shot Instance Segmentation (FSIS) requires models to detect and segment novel classes with limited several support examples. In this work, we explore a simple yet unified solution for FSIS as well as its incremental variants, and introduce a new framework named Reference Twice (RefT) to fully explore the relationship between support/query features based on a Transformer-like framework. Our key insights are two folds: Firstly, with the aid of support masks, we can generate dynamic class centers more appropriately to re-weight query features. Secondly, we find that support object queries have already encoded key factors after base training. In this way, the query features can be enhanced twice from two aspects, i.e., feature-level and instance-level. In particular, we firstly design a mask-based dynamic weighting module to enhance support features and then propose to link object queries for better calibration via cross-attention. After the above steps, the novel classes can be improved significantly over our strong baseline. Additionally, our new framework can be easily extended to incremental FSIS with minor modification. When benchmarking results on the COCO dataset for FSIS, gFSIS, and iFSIS settings, our method achieves a competitive performance compared to existing approaches across different shots, e.g., we boost nAP by noticeable +8.2/+9.4 over the current state-of-the-art FSIS method for 10/30-shot. We further demonstrate the superiority of our approach on Few Shot Object Detection. Code and model will be available.
translated by 谷歌翻译